18 research outputs found

    Constructive Type-Logical Supertagging with Self-Attention Networks

    Get PDF
    We propose a novel application of self-attention networks towards grammar induction. We present an attention-based supertagger for a refined type-logical grammar, trained on constructing types inductively. In addition to achieving a high overall type accuracy, our model is able to learn the syntax of the grammar's type system along with its denotational semantics. This lifts the closed world assumption commonly made by lexicalized grammar supertaggers, greatly enhancing its generalization potential. This is evidenced both by its adequate accuracy over sparse word types and its ability to correctly construct complex types never seen during training, which, to the best of our knowledge, was as of yet unaccomplished.Comment: REPL4NLP 4, ACL 201

    Fighting the COVID-19 Infodemic with a Holistic BERT Ensemble

    Get PDF
    This paper describes the TOKOFOU system, an ensemble model for misinformation detection tasks based on six different transformer-based pre-trained encoders, implemented in the context of the COVID-19 Infodemic Shared Task for English. We fine tune each model on each of the task's questions and aggregate their prediction scores using a majority voting approach. TOKOFOU obtains an overall F1 score of 89.7%, ranking first.Comment: 4 pages, NLP4IF 202

    A Logic-Based Framework for Natural Language Inference in Dutch

    Get PDF
    We present a framework for deriving inference relations between Dutch sentence pairs. The proposed framework relies on logic-based reasoning to produce inspectable proofs leading up to inference labels; its judgements are therefore transparent and formally verifiable. At its core, the system is powered by two λ-calculi, used as syntactic and semantic theories, respectively. Sentences are first converted to syntactic proofs and terms of the linear λ-calculus using a choice of two parsers: an Alpino-based pipeline, and Neural Proof Nets. The syntactic terms are then converted to semantic terms of the simply typed λ-calculus, via a set of hand designed type- and term-level transformations. Pairs of semantic terms are then fed to an automated theorem prover for natural logic which reasons with them while using lexical relations found in the Open Dutch WordNet. We evaluate the reasoning pipeline on the recently created Dutch natural language inference dataset, and achieve promising results, remaining only within a 1.1−3.2% performance margin to strong neural baselines. To the best of our knowledge, the reasoning pipeline is the first logic-based system for Dutch

    A Logic-Based Framework for Natural Language Inference in Dutch

    Get PDF
    We present a framework for deriving inference relations between Dutch sentence pairs. The proposed framework relies on logic-based reasoning to produce inspectable proofs leading up to inference labels; its judgements are therefore transparent and formally verifiable. At its core, the system is powered by two λ-calculi, used as syntactic and semantic theories, respectively. Sentences are first converted to syntactic proofs and terms of the linear λ-calculus using a choice of two parsers: an Alpino-based pipeline, and Neural Proof Nets. The syntactic terms are then converted to semantic terms of the simply typed λ-calculus, via a set of hand designed type- and term-level transformations. Pairs of semantic terms are then fed to an automated theorem prover for natural logic which reasons with them while using lexical relations found in the Open Dutch WordNet. We evaluate the reasoning pipeline on the recently created Dutch natural language inference dataset, and achieve promising results, remaining only within a 1.1−3.2% performance margin to strong neural baselines. To the best of our knowledge, the reasoning pipeline is the first logic-based system for Dutch

    Discontinuous Constituency and BERT: A Case Study of Dutch

    Get PDF
    In this paper, we set out to quantify the syntactic capacity of BERT in the evaluation regime of non-context free patterns, as occurring in Dutch. We devise a test suite based on a mildly context-sensitive formalism, from which we derive grammars that capture the linguistic phenomena of control verb nesting and verb raising. The grammars, paired with a small lexicon, provide us with a large collection of naturalistic utterances, annotated with verb-subject pairings, that serve as the evaluation test bed for an attention-based span selection probe. Our results, backed by extensive analysis, suggest that the models investigated fail in the implicit acquisition of the dependencies examined

    Neural Proof Nets

    Get PDF
    Linear logic and the linear {\lambda}-calculus have a long standing tradition in the study of natural language form and meaning. Among the proof calculi of linear logic, proof nets are of particular interest, offering an attractive geometric representation of derivations that is unburdened by the bureaucratic complications of conventional prooftheoretic formats. Building on recent advances in set-theoretic learning, we propose a neural variant of proof nets based on Sinkhorn networks, which allows us to translate parsing as the problem of extracting syntactic primitives and permuting them into alignment. Our methodology induces a batch-efficient, end-to-end differentiable architecture that actualizes a formally grounded yet highly efficient neuro-symbolic parser. We test our approach on {\AE}Thel, a dataset of type-logical derivations for written Dutch, where it manages to correctly transcribe raw text sentences into proofs and terms of the linear {\lambda}-calculus with an accuracy of as high as 70%.Comment: 14 pages, CoNLL202

    OYXOY: A Modern NLP Test Suite for Modern Greek

    Full text link
    This paper serves as a foundational step towards the development of a linguistically motivated and technically relevant evaluation suite for Greek NLP. We initiate this endeavor by introducing four expert-verified evaluation tasks, specifically targeted at natural language inference, word sense disambiguation (through example comparison or sense selection) and metaphor detection. More than language-adapted replicas of existing tasks, we contribute two innovations which will resonate with the broader resource and evaluation community. Firstly, our inference dataset is the first of its kind, marking not just \textit{one}, but rather \textit{all} possible inference labels, accounting for possible shifts due to e.g. ambiguity or polysemy. Secondly, we demonstrate a cost-efficient method to obtain datasets for under-resourced languages. Using ChatGPT as a language-neutral parser, we transform the Dictionary of Standard Modern Greek into a structured format, from which we derive the other three tasks through simple projections. Alongside each task, we conduct experiments using currently available state of the art machinery. Our experimental baselines affirm the challenging nature of our tasks and highlight the need for expedited progress in order for the Greek NLP ecosystem to keep pace with contemporary mainstream research

    Dependency as Modality, Parsing as Permutation: A Neurosymbolic Perspective on Categorial Grammars

    No full text
    Type systems (in the form of categorial grammars) are the front runners in the quest for a formally elegant, computationally attractive and flexible theory of linguistic form and meaning. Words enact typed constants, and interact with one another via means of grammatical rules enacted by type inferences, composing larger phrases in the process. The end result is at the same time a parse, a proof and a program, bridging the seemingly disparate fields of linguistics, formal logics and computer science. The transition from form to meaning is traditionally handled via a series of homomorphisms that simplify nuances of the syntactic calculus to move towards a uniform semantic calculus. Alluring as this might be, it poses pragmatic considerations. For the setup to work on the semantic level, one has no choice but to start from the hardest part, namely natural language syntax. Phenomena like movement, word-order variation, discontinuities, etc. require careful treatment that needs to be both general enough to encompass the full range of grammatical utterances, yet strict enough to ward off ungrammatical derivations. This thesis takes an operational shortcut in targeting a ``deeper'' calculus of grammatical composition, engaging only minimally with surface syntax. Where previously functional functional syntactic types would be position-conscious, requiring their arguments in predetermined positions upon a binary tree, they are now agnostic to both tree structure and sequential order, alleviating the need for syntactic refinements. This simplification comes at the cost of a misalignment between provability and grammaticality: the laxer semantic calculus permits more proofs than linguistically desired. To circumvent this underspecification, the thesis takes a step away from the established norm, proposing the incorporation of unary type operators extending the function-argument axis with grammatical role labels. The new calculus produces mixed unary/n-ary trees, each unary tree denoting a dependency domain, and each n-ary tree denoting the phrases which together form that domain. Although still underspecified, these structures now subsume non-projective labeled dependency trees. More than that, they have their roots set firmly in type theory, allowing meaningful semantic interpretation. On more practical grounds and in order to investigate the formalism's adequacy, an extraction algorithm is employed to convert syntactic analyses of sentences (represented as dependency graphs) into proofs of the target logic. This gives rise to a large proofbank, a collection of sentences paired with tectogrammatic theorems and their corresponding programs, and an elaborate type lexicon, providing type assignments to one million lexical tokens within a linguistic context. The proofbank and the lexicon find use as training data in the design of a neurosymbolic proof search system, able to efficiently navigate the logic's theorem space. The system consists of three components. Component one is a supertagger responsible for assigning a type to each word — the tagger is formulated as a heterogeneous graph convolution kernel that boasts state-of-the-art accuracy. Rather than produce asignments in the form of conditional probabilities over a predefined vocabulary, it instead constructs types dynamicaly. As such, it is unconstrained by data sparsity, generalizing well to rare assignments and producing correct assignments for types never seen during training. Component two is a neural permutation module that exploits the linearity constraint of the logic in order to simplify proof search as optimal transport learning, associating resources (conditional validities) to the processes that require them (conditions). This allows for a parallel and easily optimizable implementation, unobstructed by the structure manipulation found in conventional parsers. Component three is the type system itself, responsible for navigating the produced structures and asserting their well-formedness. Results suggest performance superior to established baselines across categorial formalisms, despite the ambiguity inherent to the logic
    corecore